Categorical Foundations of Gradient-Based Learning

نویسندگان

چکیده

Abstract We propose a categorical semantics of gradient-based machine learning algorithms in terms lenses, parametric maps, and reverse derivative categories. This foundation provides powerful explanatory unifying framework: it encompasses variety gradient descent such as ADAM, AdaGrad, Nesterov momentum, well loss functions MSE Softmax cross-entropy, shedding new light on their similarities differences. Our approach to has examples generalising beyond the familiar continuous domains (modelled categories smooth maps) can be realized discrete setting boolean circuits. Finally, we demonstrate practical significance our framework with an implementation Python.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Solomon Feferman Categorical Foundations and Foundations of Category Theory*

This paper is divided'into two parts. Part I deals briefly with the thesis that category theory (or something like it) should provide the proper foundations of mathematics, in preference to current foundational schemes. An opposite view is argued here on the grounds that the notions of operation and collection are prior to all structural notions. However, no position is taken as to whether such...

متن کامل

Foundations of Game-Based Learning

s on human factors in computing systems (pp. 1885–1890). New

متن کامل

Reflections on a categorical foundations of mathematics

We examine Gödel’s completeness and incompleteness theorems for higher order arithmetic from a categorical point of view. The former says that a proposition is provable if and only if it is true in all models, which we take to be local toposes, i.e. Lawvere’s elementary toposes in which the terminal object is a nontrivial indecomposable projective. The incompleteness theorem showed that, in the...

متن کامل

Categorical Foundations of Distributed Graph Transformation

A distributed graph (N, D) consists of a network graph N and a commutative diagram D over the scheme N which associates local graphs D(ni) and graph morphisms D(e) : D(n1) → D(n2) to nodes n1, n2 and edges e : n1 → n2 in N . Although there are several interesting applications of distributed graphs and transformations, even the basic pushout constructions for the double pushout approach of distr...

متن کامل

Gradient - Based Learning Applied

| Multilayer Neural Networks trained with the backpropa-gation algorithm constitute the best example of a successful Gradient-Based Learning technique. Given an appropriate network architecture, Gradient-Based Learning algorithms can be used to synthesize a complex decision surface that can classify high-dimensional patterns such as handwritten characters , with minimal preprocessing. This pape...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Lecture Notes in Computer Science

سال: 2022

ISSN: ['1611-3349', '0302-9743']

DOI: https://doi.org/10.1007/978-3-030-99336-8_1